runtime._type.ptrdata (field)
62 uses
runtime (current package)
cgocall.go#L446: if t.ptrdata == 0 || p == nil {
cgocall.go#L509: if st.elem.ptrdata == 0 {
cgocall.go#L534: if f.typ.ptrdata == 0 {
cgocheck.go#L67: if typ.ptrdata == 0 {
cgocheck.go#L87: if typ.ptrdata == 0 {
cgocheck.go#L110: if typ.ptrdata <= off {
cgocheck.go#L113: if ptrdataSize := typ.ptrdata - off; size > ptrdataSize {
cgocheck.go#L207: if typ.ptrdata == 0 {
cgocheck.go#L212: if typ.ptrdata <= off {
cgocheck.go#L215: if ptrdataSize := typ.ptrdata - off; size > ptrdataSize {
chan.go#L99: case elem.ptrdata == 0:
checkptr.go#L19: if elem.ptrdata != 0 && uintptr(p)&(uintptr(elem.align)-1) != 0 {
heapdump.go#L209: dumpbool(t.kind&kindDirectIface == 0 || t.ptrdata != 0)
malloc.go#L995: noscan := typ == nil || typ.ptrdata == 0
malloc.go#L1122: if typ.ptrdata != 0 {
malloc.go#L1123: scanSize = dataSize - typ.size + typ.ptrdata
malloc.go#L1126: scanSize = typ.ptrdata
map.go#L266: if t.bucket.ptrdata == 0 {
map.go#L370: if t.bucket.ptrdata != 0 {
map.go#L751: } else if t.key.ptrdata != 0 {
map.go#L757: } else if t.elem.ptrdata != 0 {
map.go#L834: if t.bucket.ptrdata == 0 {
map.go#L1249: if h.flags&oldIterator == 0 && t.bucket.ptrdata != 0 {
map_fast32.go#L305: if goarch.PtrSize == 4 && t.key.ptrdata != 0 {
map_fast32.go#L311: if t.elem.ptrdata != 0 {
map_fast32.go#L431: if goarch.PtrSize == 4 && t.key.ptrdata != 0 && writeBarrier.enabled {
map_fast32.go#L449: if h.flags&oldIterator == 0 && t.bucket.ptrdata != 0 {
map_fast64.go#L303: if t.key.ptrdata != 0 {
map_fast64.go#L313: if t.elem.ptrdata != 0 {
map_fast64.go#L433: if t.key.ptrdata != 0 && writeBarrier.enabled {
map_fast64.go#L457: if h.flags&oldIterator == 0 && t.bucket.ptrdata != 0 {
map_faststr.go#L339: if t.elem.ptrdata != 0 {
map_faststr.go#L472: if h.flags&oldIterator == 0 && t.bucket.ptrdata != 0 {
mbarrier.go#L161: if writeBarrier.needed && typ.ptrdata != 0 {
mbarrier.go#L162: bulkBarrierPreWrite(uintptr(dst), uintptr(src), typ.ptrdata)
mbarrier.go#L204: if writeBarrier.needed && typ.ptrdata > off && size >= goarch.PtrSize {
mbarrier.go#L209: if poff := typ.ptrdata - off; pwsize > poff {
mbarrier.go#L232: if writeBarrier.needed && typ != nil && typ.ptrdata != 0 && size >= goarch.PtrSize {
mbarrier.go#L288: pwsize := size - typ.size + typ.ptrdata
mbarrier.go#L299: if elemType.ptrdata == 0 {
mbarrier.go#L315: if writeBarrier.needed && typ.ptrdata != 0 {
mbarrier.go#L316: bulkBarrierPreWrite(uintptr(ptr), 0, typ.ptrdata)
mbarrier.go#L328: if writeBarrier.needed && typ.ptrdata != 0 {
mbitmap.go#L733: for i := uintptr(0); i < typ.ptrdata; i += goarch.PtrSize {
mbitmap.go#L906: hb |= bitScanAll & ((bitScan << (typ.ptrdata / goarch.PtrSize)) - 1)
mbitmap.go#L1021: heapBitsSetTypeGCProg(h, typ.ptrdata, typ.size, dataSize, size, addb(typ.gcdata, 4))
mbitmap.go#L1077: if typ.ptrdata/goarch.PtrSize <= maxBits {
mbitmap.go#L1088: nb = typ.ptrdata / goarch.PtrSize
mbitmap.go#L1123: n := (typ.ptrdata/goarch.PtrSize+7)/8 - 1
mbitmap.go#L1136: nw = typ.ptrdata / goarch.PtrSize
mbitmap.go#L1141: nw = ((dataSize/typ.size-1)*typ.size + typ.ptrdata) / goarch.PtrSize
mbitmap.go#L1419: print("typ.size=", typ.size, " typ.ptrdata=", typ.ptrdata, " dataSize=", dataSize, " size=", size, "\n")
mbitmap.go#L1430: nptr := typ.ptrdata / goarch.PtrSize
mbitmap.go#L1433: totalptr := ((count-1)*typ.size + typ.ptrdata) / goarch.PtrSize
mbitmap.go#L1452: print("typ.size=", typ.size, " typ.ptrdata=", typ.ptrdata, " dataSize=", dataSize, " size=", size, "\n")
mbitmap.go#L1958: nptr := typ.ptrdata / goarch.PtrSize
mfinal.go#L376: if ot.elem == nil || ot.elem.ptrdata != 0 || ot.elem.size >= maxTinySize {
slice.go#L56: if et.ptrdata == 0 {
slice.go#L271: if et.ptrdata == 0 {
slice.go#L282: bulkBarrierPreWriteSrcOnly(uintptr(p), uintptr(old.array), lenmem-et.size+et.ptrdata)
stack.go#L1382: _ptrdata: int32(abiRegArgsType.ptrdata),
type.go#L36: ptrdata uintptr // size of memory prefix holding all pointers